Fortran 90d Intrinsic Functions on Distributed Memory Machines: Implementation and Scalability Fortran 90d Intrinsic Functions on Distributed Memory Machines: Implementation and Scalability
نویسندگان
چکیده
We are developing a Fortran 90D compiler, which converts Fortran 90D code into Fortran 77 plus message passing node programs for distributed memory machines. This paper presents the implementation and performance results of Fortran 90D intrinsic functions on the Intel iPSC/860 hypercube. Our implementation is portable and scalable.
منابع مشابه
Supporting Irregular Distributions Using S DatacParallel Languages
12 0 n distributed-memory machines, large data arrays need to be partitioned between local processor memories. These partitioned data arrays are called distributed arrays. Many applications can be efficiently implemented by using simple schemes for mapping distributed arrays. One such scheme is BLOCK distribution, which divides an array into contiguous, equal-sized subarrays and assigns each su...
متن کاملCompiling Distribution Directives in a FORTRAN 90D Compiler
Data Partitioning and mapping is one of the most important steps of in writing a parallel program; especially data parallel one. Recently, Fortran D, and subsequently, High Performance Fortran (HPF) have been proposed to allow users to specify data distributions and alignments for arrays in programs. This paper presents the design of a Fortran 90D compiler that takes a Fortran 90D program as in...
متن کاملA Compilation Approach for Fortran 90D/ HPF Compilers
This paper describes a compilation approach for a Fortran 90D/HPF compiler, a source-to-source parallel compiler for distributed memory systems. Diierent from Fortran 77 parallelizing compilers, a Fortran90D/HPF compiler does not parallelize sequential constructs. Only par-allelism expressed by Fortran 90D/HPF parallel constructs is exploited. The methodoly of parallelizing Fortran programs suc...
متن کاملProtein Folding with Python on Supercomputers
Today’s supercomputers have hundreds of thousands of compute cores and this number is likely to grow. Many of today’s algorithms will have to be rethought to take advantage of such large systems. New algorithms must provide fine grained parallelism and excellent scalability. Python offers good support for numerical libraries and offers bindings to MPI that can be used to develop parallel algori...
متن کاملAn Integrated Runtime and Compile-time Approach for Parallelizing Structured and Block Structured Applications an Integrated Runtime and Compile-time Approach for Parallelizing Structured and Block Structured Applications
Scientiic and engineering applications often involve structured meshes. These meshes may be nested (for multigrid codes) and/or irregularly coupled (called multiblock or irregularly coupled regular mesh problems). In this paper, we present a combined runtime and compile-time approach for parallelizing these applications on distributed memory parallel machines in an eecient and machine-independe...
متن کامل